var runtime.debug

121 uses

	runtime (current package)
		arena.go#L816: 	if debug.malloc {
		cgocall.go#L431: 	if debug.dataindependenttiming == 1 && gp.m.isextra {
		cgocall.go#L452: 	if debug.dataindependenttiming == 1 && !ditAlreadySet {
		cgocall.go#L538: 	if !goexperiment.CgoCheck2 && debug.cgocheck == 0 {
		cgocall.go#L789: 	if !goexperiment.CgoCheck2 && debug.cgocheck == 0 {
		chan.go#L819: 	async := debug.asynctimerchan.Load() != 0
		chan.go#L837: 		async := debug.asynctimerchan.Load() != 0
		malloc.go#L1028: 	if debug.malloc {
		malloc.go#L1092: 	if debug.malloc {
		malloc.go#L1615: 	if debug.sbrk != 0 {
		mbitmap.go#L1276: 		if (GOARCH == "amd64" || GOARCH == "arm64") && p == clobberdeadPtr && debug.invalidptr != 0 {
		mbitmap.go#L1295: 		if debug.invalidptr != 0 {
		mcleanup.go#L89: 	if debug.sbrk != 0 {
		mem_linux.go#L50: 	if debug.madvdontneed != 0 && advise != madviseUnsupported {
		mem_linux.go#L76: 	if debug.harddecommit > 0 {
		mem_linux.go#L85: 	if debug.harddecommit > 0 {
		mem_linux.go#L179: 	if debug.disablethp != 0 {
		mfinal.go#L451: 	if debug.sbrk != 0 {
		mgc.go#L680: 	if debug.gcstoptheworld == 1 {
		mgc.go#L682: 	} else if debug.gcstoptheworld == 2 {
		mgc.go#L1054: 		if debug.gccheckmark > 0 {
		mgc.go#L1237: 	if debug.gctrace > 0 {
		mgc.go#L1596: 	if debug.gccheckmark > 0 {
		mgc.go#L1619: 		if debug.gccheckmark > 0 {
		mgcmark.go#L1598: 		if debug.gccheckmark > 0 && span.isFree(objIndex) {
		mgcpacer.go#L420: 	if debug.gcstoptheworld > 0 {
		mgcpacer.go#L456: 	if debug.gcpacertrace > 0 {
		mgcpacer.go#L674: 	if debug.gcpacertrace > 0 {
		mgcsweep.go#L175: 			if debug.gcpacertrace > 0 {
		mgcsweep.go#L420: 		if debug.scavtrace > 0 {
		mgcsweep.go#L610: 	if traceAllocFreeEnabled() || debug.clobberfree != 0 || raceenabled || msanenabled || asanenabled {
		mgcsweep.go#L624: 				if debug.clobberfree != 0 {
		mgcsweep.go#L814: 			if debug.efence > 0 {
		mheap.go#L1695: 	if debug.scavtrace > 0 {
		mprof.go#L447: 	nstk := callers(5, mp.profStack[:debug.profstackdepth])
		mprof.go#L534: 	if debug.profstackdepth == 0 {
		mprof.go#L782: 	if debug.profstackdepth == 0 {
		mprof.go#L805: 	if debug.runtimeContentionStacks.Load() == 0 {
		mprof.go#L833: 	nstk := int(debug.profstackdepth)
		panic.go#L744: 		if debug.panicnil.Load() != 1 {
		panic.go#L1373: 		if debug.schedtrace > 0 || debug.scheddetail > 0 {
		preempt.go#L223: 			if preemptMSupported && debug.asyncpreemptoff == 0 && needAsync {
		proc.go#L196: 	if debug.inittrace != 0 {
		proc.go#L350: 		if debug.gctrace > 0 {
		proc.go#L986: 	if debug.profstackdepth == 0 {
		proc.go#L1005: 	return make([]uintptr, 1+maxSkip+debug.profstackdepth)
		proc.go#L1010: func makeProfStack() []uintptr { return make([]uintptr, debug.profstackdepth) }
		proc.go#L1075: 	if debug.dontfreezetheworld > 0 {
		proc.go#L1860: 	if debug.dataindependenttiming == 1 {
		proc.go#L4040: 	if debug.dontfreezetheworld > 0 && freezing.Load() {
		proc.go#L5168: 	if debug.tracebackancestors <= 0 || callergp.goid == 0 {
		proc.go#L5176: 	if n > debug.tracebackancestors {
		proc.go#L5177: 		n = debug.tracebackancestors
		proc.go#L6147: 		if debug.schedtrace <= 0 && (sched.gcwaiting.Load() || sched.npidle.Load() == gomaxprocs) {
		proc.go#L6249: 		if debug.schedtrace > 0 && lasttrace+int64(debug.schedtrace)*1000000 <= now {
		proc.go#L6251: 			schedtrace(debug.scheddetail > 0)
		proc.go#L6390: 	if preemptMSupported && debug.asyncpreemptoff == 0 {
		runtime1.go#L309: var debug struct {
		runtime1.go#L366: 	{name: "adaptivestackstart", value: &debug.adaptivestackstart},
		runtime1.go#L367: 	{name: "asyncpreemptoff", value: &debug.asyncpreemptoff},
		runtime1.go#L368: 	{name: "asynctimerchan", atomic: &debug.asynctimerchan},
		runtime1.go#L369: 	{name: "cgocheck", value: &debug.cgocheck},
		runtime1.go#L370: 	{name: "clobberfree", value: &debug.clobberfree},
		runtime1.go#L371: 	{name: "dataindependenttiming", value: &debug.dataindependenttiming},
		runtime1.go#L372: 	{name: "disablethp", value: &debug.disablethp},
		runtime1.go#L373: 	{name: "dontfreezetheworld", value: &debug.dontfreezetheworld},
		runtime1.go#L374: 	{name: "efence", value: &debug.efence},
		runtime1.go#L375: 	{name: "gccheckmark", value: &debug.gccheckmark},
		runtime1.go#L376: 	{name: "gcpacertrace", value: &debug.gcpacertrace},
		runtime1.go#L377: 	{name: "gcshrinkstackoff", value: &debug.gcshrinkstackoff},
		runtime1.go#L378: 	{name: "gcstoptheworld", value: &debug.gcstoptheworld},
		runtime1.go#L379: 	{name: "gctrace", value: &debug.gctrace},
		runtime1.go#L380: 	{name: "harddecommit", value: &debug.harddecommit},
		runtime1.go#L381: 	{name: "inittrace", value: &debug.inittrace},
		runtime1.go#L382: 	{name: "invalidptr", value: &debug.invalidptr},
		runtime1.go#L383: 	{name: "madvdontneed", value: &debug.madvdontneed},
		runtime1.go#L384: 	{name: "panicnil", atomic: &debug.panicnil},
		runtime1.go#L385: 	{name: "profstackdepth", value: &debug.profstackdepth, def: 128},
		runtime1.go#L386: 	{name: "runtimecontentionstacks", atomic: &debug.runtimeContentionStacks},
		runtime1.go#L387: 	{name: "sbrk", value: &debug.sbrk},
		runtime1.go#L388: 	{name: "scavtrace", value: &debug.scavtrace},
		runtime1.go#L389: 	{name: "scheddetail", value: &debug.scheddetail},
		runtime1.go#L390: 	{name: "schedtrace", value: &debug.schedtrace},
		runtime1.go#L391: 	{name: "traceadvanceperiod", value: &debug.traceadvanceperiod},
		runtime1.go#L392: 	{name: "traceallocfree", atomic: &debug.traceallocfree},
		runtime1.go#L393: 	{name: "tracecheckstackownership", value: &debug.traceCheckStackOwnership},
		runtime1.go#L394: 	{name: "tracebackancestors", value: &debug.tracebackancestors},
		runtime1.go#L395: 	{name: "tracefpunwindoff", value: &debug.tracefpunwindoff},
		runtime1.go#L400: 	debug.cgocheck = 1
		runtime1.go#L401: 	debug.invalidptr = 1
		runtime1.go#L402: 	debug.adaptivestackstart = 1 // set this to 0 to turn larger initial goroutine stacks off
		runtime1.go#L412: 		debug.madvdontneed = 1
		runtime1.go#L414: 	debug.traceadvanceperiod = defaultTraceAdvancePeriod
		runtime1.go#L440: 	debug.malloc = (debug.inittrace | debug.sbrk) != 0
		runtime1.go#L441: 	debug.profstackdepth = min(debug.profstackdepth, maxProfStackDepth)
		runtime1.go#L530: 	if debug.cgocheck > 1 {
		signal_unix.go#L447: 		if sig == sigPreempt && preemptMSupported && debug.asyncpreemptoff == 0 {
		signal_unix.go#L685: 	if sig == sigPreempt && debug.asyncpreemptoff == 0 && !delayedSignal {
		signal_unix.go#L1391: 	if sig == sigPreempt && preemptMSupported && debug.asyncpreemptoff == 0 {
		stack.go#L351: 	if debug.efence != 0 || stackFromSystem != 0 {
		stack.go#L460: 	if debug.efence != 0 || stackFromSystem != 0 {
		stack.go#L461: 		if debug.efence != 0 || stackFaultOnFree != 0 {
		stack.go#L635: 			if f.valid() && 0 < p && p < minLegalPointer && debug.invalidptr != 0 {
		stack.go#L1206: 	if debug.gcshrinkstackoff > 0 {
		stack.go#L1319: 	if debug.adaptivestackstart == 0 {
		synctest.go#L155: 	if debug.asynctimerchan.Load() != 0 {
		time.go#L473: 	async := debug.asynctimerchan.Load() != 0
		time.go#L554: 	async := debug.asynctimerchan.Load() != 0
		time.go#L1108: 	async := debug.asynctimerchan.Load() != 0
		trace.go#L254: 	if debug.traceallocfree.Load() != 0 {
		trace.go#L256: 		trace.debugMalloc = debug.malloc
		trace.go#L258: 		debug.malloc = true
		trace.go#L628: 			debug.malloc = trace.debugMalloc
		trace.go#L960: 			s.timer.sleep(int64(debug.traceadvanceperiod))
		tracestack.go#L49: 	if debug.traceCheckStackOwnership != 0 && gp != nil {
		tracestack.go#L248: 	return debug.tracefpunwindoff != 0 || (goarch.ArchFamily != goarch.AMD64 && goarch.ArchFamily != goarch.ARM64)